贝叶斯工作流程通常需要引入滋扰参数,但对于核心科学建模,需要访问边缘后部密度。在这项工作中,我们使用掩盖的自回归流量和内核密度估计器封装边缘后部,使我们能够计算边际kullback-leibler脱离器和边缘贝叶斯模型尺寸,此外还可以生成样品和计算边际对数概率。我们将其应用于暗能量调查的局部宇宙学示例和全局21cm信号实验。除了计算边缘贝叶斯统计数据外,这项工作对于在贝叶斯实验设计,复杂的先验建模和似然仿真中进一步应用也很重要。该技术可在PIP可容纳的代码人造黄油中公开获得。
translated by 谷歌翻译
This paper is a technical overview of DeepMind and Google's recent work on reinforcement learning for controlling commercial cooling systems. Building on expertise that began with cooling Google's data centers more efficiently, we recently conducted live experiments on two real-world facilities in partnership with Trane Technologies, a building management system provider. These live experiments had a variety of challenges in areas such as evaluation, learning from offline data, and constraint satisfaction. Our paper describes these challenges in the hope that awareness of them will benefit future applied RL work. We also describe the way we adapted our RL system to deal with these challenges, resulting in energy savings of approximately 9% and 13% respectively at the two live experiment sites.
translated by 谷歌翻译
Large language models (LLMs) have been shown to be able to perform new tasks based on a few demonstrations or natural language instructions. While these capabilities have led to widespread adoption, most LLMs are developed by resource-rich organizations and are frequently kept from the public. As a step towards democratizing this powerful technology, we present BLOOM, a 176B-parameter open-access language model designed and built thanks to a collaboration of hundreds of researchers. BLOOM is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total). We find that BLOOM achieves competitive performance on a wide variety of benchmarks, with stronger results after undergoing multitask prompted finetuning. To facilitate future research and applications using LLMs, we publicly release our models and code under the Responsible AI License.
translated by 谷歌翻译
Deep learning-based pose estimation algorithms can successfully estimate the pose of objects in an image, especially in the field of color images. 6D Object pose estimation based on deep learning models for X-ray images often use custom architectures that employ extensive CAD models and simulated data for training purposes. Recent RGB-based methods opt to solve pose estimation problems using small datasets, making them more attractive for the X-ray domain where medical data is scarcely available. We refine an existing RGB-based model (SingleShotPose) to estimate the 6D pose of a marked cube from grayscale X-ray images by creating a generic solution trained on only real X-ray data and adjusted for X-ray acquisition geometry. The model regresses 2D control points and calculates the pose through 2D/3D correspondences using Perspective-n-Point(PnP), allowing a single trained model to be used across all supporting cone-beam-based X-ray geometries. Since modern X-ray systems continuously adjust acquisition parameters during a procedure, it is essential for such a pose estimation network to consider these parameters in order to be deployed successfully and find a real use case. With a 5-cm/5-degree accuracy of 93% and an average 3D rotation error of 2.2 degrees, the results of the proposed approach are comparable with state-of-the-art alternatives, while requiring significantly less real training examples and being applicable in real-time applications.
translated by 谷歌翻译
通用数据模型解决了标准化电子健康记录(EHR)数据的许多挑战,但无法将其集成深度表型所需的资源。开放的生物学和生物医学本体论(OBO)铸造本体论提供了可用于生物学知识的语义计算表示,并能够整合多种生物医学数据。但是,将EHR数据映射到OBO Foundry本体论需要大量的手动策展和域专业知识。我们介绍了一个框架,用于将观察性医学成果合作伙伴关系(OMOP)标准词汇介绍给OBO铸造本体。使用此框架,我们制作了92,367条条件,8,615种药物成分和10,673个测量结果的映射。域专家验证了映射准确性,并且在24家医院进行检查时,映射覆盖了99%的条件和药物成分和68%的测量结果。最后,我们证明OMOP2OBO映射可以帮助系统地识别可能受益于基因检测的未诊断罕见病患者。
translated by 谷歌翻译
生成与历史数据具有相似分布和依赖性的电力系统对于系统规划和安全评估的任务至关重要,尤其是在历史数据不足的情况下。在本文中,我们根据有条件的变异自动编码器(CVAE)神经网络体系结构描述了工业和商业客户负载概况的生成模型,由于此类配置文件的高度可变性质,该模型具有挑战性。生成的上下文负载轮廓是在一年中的一个月进行的,并与电网进行了典型的电力交换。此外,世代的质量在视觉和统计上都经过评估。实验结果表明,我们提出的CVAE模型可以捕获历史负载谱的时间特征,并以满意的单变量分布和多元依赖性生成“现实”数据。
translated by 谷歌翻译
从有限的资源中获得最大收益可以进步自然语言处理(NLP)研究和实践,同时保守资源。这些资源可能是数据,时间,存储或能源。NLP的最新工作从缩放率产生了有趣的结果。但是,仅使用比例来改善结果意味着资源消耗也会扩展。这种关系激发了对有效方法的研究,这些方法需要更少的资源才能获得相似的结果。这项调查涉及NLP效率的方法和发现,旨在指导该领域的新研究人员并激发新方法的发展。
translated by 谷歌翻译
现代机器学习任务通常不仅需要考虑一个目标,而且需要考虑多个目标。例如,除了预测质量外,这可能是学识渊博的模型或其任何组合的效率,稳健性或公平性。多目标学习为处理此类问题提供了自然框架,而无需提交早期权衡。令人惊讶的是,到目前为止,统计学习理论几乎没有深入了解多目标学习的概括属性。在这项工作中,我们采取了第一步来填补这一空白:我们为多目标设置建立了基础概括范围,以及通过标量化学习的概括和超级界限。我们还提供了对真实目标的帕累托最佳集合与他们从训练数据中经验近似的帕累托(Pareto)最佳选择之间的关系的第一个理论分析。特别是,我们表现出令人惊讶的不对称性:所有帕累托最佳的解决方案都可以通过经验上的帕累托(Pareto)优势近似,但反之亦然。
translated by 谷歌翻译
机器学习(ML)是一种在车辆互联网(IOV)上培训预测模型的分布式方法,以实现智能公共交通。由于交通状况会随着时间而变化,因此必须连续有效地更新流量流动和乘客等待时间的ML模型。联合学习(FL)是一种分布式机器学习方案,允许车辆接收连续的模型更新,而无需将原始数据上传到云中并等待培训模型。但是,由于车辆在公共场所旅行以来,智能公共交通中FL容易受到中毒或DDOS攻击的影响。此外,由于设备异质性和不平衡数据分布,同步聚合策略在聚集之前从特定车辆中收集本地模型的同步聚合策略效率低下。尽管有异步联合学习(AFL)方案是通过收到本地模型来提高效率的,但陈旧的本地模型仍然不合理地加权,导致学习绩效不佳。为了实现更明智的公共交通,本文提供了一个基于动态缩放系数(DBAFL)的基于区块链的异步联合学习方案。具体而言,基于委员会的新型共识算法用于区块链,以最低的时间成本提高了可靠性。同时,设计的动态缩放系数允许AFL为陈旧的本地模型分配合理的重量。在异质设备上进行的广泛实验验证了DBAFL的学习效果,效率和可靠性优于外观的实验。
translated by 谷歌翻译
黑色素瘤是一种严重的皮肤癌,在后期阶段高死亡率。幸运的是,当早期发现时,黑色素瘤的预后是有希望的,恶性黑色素瘤的发病率相对较低。结果,数据集严重不平衡,这使培训当前的最新监督分类AI模型变得复杂。我们建议使用生成模型来学习良性数据分布,并通过密度估计检测出分布(OOD)恶性图像。标准化流(NFS)是OOD检测的理想候选者,因为它们可以计算精确的可能性。然而,它们的感应偏见对明显的图形特征而不是语义上下文障碍障碍的OOD检测。在这项工作中,我们旨在将这些偏见与黑色素瘤的领域水平知识一起使用,以改善基于可能性的OOD检测恶性图像。我们令人鼓舞的结果表明,使用NFS检测黑色素瘤的可能性。我们通过使用基于小波的NFS,在接收器工作特性的曲线下,面积增加了9%。该模型需要较少的参数,以使其更适用于边缘设备。拟议的方法可以帮助医学专家诊断出皮肤癌患者并不断提高存活率。此外,这项研究为肿瘤学领域的其他领域铺平了道路,具有类似的数据不平衡问题\ footNote {代码可用:
translated by 谷歌翻译